Automatic Selection of Split Criterion during Tree Growing Based on Node Location
نویسنده
چکیده
Typically, decision tree construction algorithms apply a single \goodness of split" criterion to form each test node of the tree. It is a hypothesis of this research that better results can be obtained if during tree construction one applies a split criterion suited to the \location" of the test node in the tree. Speciically, given the objective of maximizing predictive accuracy, test nodes near the root of the tree should be chosen using a measure based on information theory, whereas test nodes closer to the leaves of the pruned tree should be chosen to maximize classii-cation accuracy on the training set. The results of an empirical evaluation illustrate that adapting the split criterion to node location can improve classiication performance. 1 DECISION TREE CONSTRUCTION A decision tree is either a leaf node containing a classi-cation or an attribute test, with for each value of the attribute, a branch to a decision tree. To classify an instance using a decision tree, one starts at the root node and nds the branch corresponding to the value of the test attribute observed in the instance. This process repeats at the subtree rooted at that branch until a leaf node is reached. The resulting classiication is the class label of the leaf. The objective of a decision tree construction algorithm is to create a tree such that the classiication accuracy of the tree when applied to previously unobserved instances is maximized (hereafter the predictive accuracy). Other criteria such as tree size and tree under-standability may also be of interest. For domains in which the cost of misclassifying instances is not uniform , the objective is to nd the tree such that misclas-siication cost (given by a misclassiication cost matrix) for previously unobserved instances is minimized. One well-known approach to constructing a decision tree is to grow a tree until each of the terminal nodes (leaves) contain instances from a single class and then prune back the tree with the goal of nding the sub-tree with the lowest misclassiication rate on previously unobserved instances (Breiman, Friedman, Olshen & Stone, 1984; Quinlan, 1986). During tree growing, at each node, one wants to select a test that best divides the instances into their classes. There are many diier-ent criteria that can be used to judge the \goodness of a split"; the most common appear in the form of an entropy or impurity measure. and compare splitting criteria. …
منابع مشابه
Attribute Dependencies, Understandability and Split Selection in Tree Based Models
The attributes’ interdependencies have strong effect on understandability of tree based models. If strong dependencies between the attributes are not recognized and these attributes are not used as splits near the root of the tree this causes node replications in lower levels of the tree, blurs the description of dependencies and also might cause drop of accuracy. If Relief family of algorithms...
متن کاملAttribute dependencies, understandability and split selection in tree based models
The attributes’ interdependencies have strong effect on understandability of tree based models. If strong dependencies between the attributes are not recognized and these attributes are not used as splits near the root of the tree this causes node replications in lower levels of the tree, blurs the description of dependencies and also might cause drop of accuracy. If Relief family of algorithms...
متن کاملAutomatic QoS-aware Web Services Composition based on Set-Cover Problem
By definition, web-services composition works on developing merely optimum coordination among a number of available web-services to provide a new composed web-service intended to satisfy some users requirements for which a single web service is not (good) enough. In this article, the formulation of the automatic web-services composition is proposed as several set-cover problems and an approxima...
متن کاملInduction of Multivariate Decision Trees by Using Dipolar Criteria
A new approach to the induction of multivariate decision trees is proposed. A linear decision function (hyper-plane) is used at each non-terminal node of a binary tree for splitting the data. The search strategy is based on the dipolar criterion functions and exploits the basis exchange algorithm as an optimization procedure. The feature selection is used to eliminate redundant and noisy featur...
متن کاملUnbiased split selection for classification trees based on the Gini Index
The Gini gain is one of the most common variable selection criteria in machine learning. We derive the exact distribution of the maximally selected Gini gain in the context of binary classification using continuous predictors by means of a combinatorial approach. This distribution provides a formal support for variable selection bias in favor of variables with a high amount of missing values wh...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 1995